Skip to content

Conversation

@armandsauzay
Copy link

Summary:
X-link: https://github.com/facebookresearch/FBGEMM/pull/2202

Add output_dtype parameter to MX4 dequantization stack to support direct
conversion to BF16/FP16, avoiding expensive FP32 intermediate step.

Differential Revision: D87826479

@meta-codesync
Copy link
Contributor

meta-codesync bot commented Dec 9, 2025

@armandsauzay has exported this pull request. If you are a Meta employee, you can view the originating Diff in D87826479.

@meta-cla meta-cla bot added the CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. label Dec 9, 2025
armandsauzay pushed a commit to armandsauzay/FBGEMM-1 that referenced this pull request Dec 9, 2025
Summary:
X-link: meta-pytorch/torchrec#3602

X-link: facebookresearch/FBGEMM#2202

Add output_dtype parameter to MX4 dequantization stack to support direct
conversion to BF16/FP16, avoiding expensive FP32 intermediate step.

Differential Revision: D87826479
armandsauzay pushed a commit to armandsauzay/torchrec that referenced this pull request Dec 10, 2025
…3602)

Summary:
X-link: pytorch/FBGEMM#5206


X-link: facebookresearch/FBGEMM#2202

Add output_dtype parameter to MX4 dequantization stack to support direct
conversion to BF16/FP16, avoiding expensive FP32 intermediate step.

Differential Revision: D87826479
armandsauzay pushed a commit to armandsauzay/FBGEMM-1 that referenced this pull request Dec 10, 2025
Summary:

X-link: meta-pytorch/torchrec#3602

X-link: facebookresearch/FBGEMM#2202

Add output_dtype parameter to MX4 dequantization stack to support direct
conversion to BF16/FP16, avoiding expensive FP32 intermediate step.

Differential Revision: D87826479
armandsauzay pushed a commit to armandsauzay/torchrec that referenced this pull request Dec 10, 2025
…3602)

Summary:
X-link: pytorch/FBGEMM#5206


X-link: facebookresearch/FBGEMM#2202

Add output_dtype parameter to MX4 dequantization stack to support direct
conversion to BF16/FP16, avoiding expensive FP32 intermediate step.

Differential Revision: D87826479
armandsauzay pushed a commit to armandsauzay/FBGEMM-1 that referenced this pull request Dec 10, 2025
Summary:

X-link: meta-pytorch/torchrec#3602

X-link: facebookresearch/FBGEMM#2202

Add output_dtype parameter to MX4 dequantization stack to support direct
conversion to BF16/FP16, avoiding expensive FP32 intermediate step.

Differential Revision: D87826479
…3602)

Summary:
X-link: pytorch/FBGEMM#5206


X-link: facebookresearch/FBGEMM#2202

Add output_dtype parameter to MX4 dequantization stack to support direct
conversion to BF16/FP16, avoiding expensive FP32 intermediate step.

Differential Revision: D87826479
armandsauzay pushed a commit to armandsauzay/FBGEMM-1 that referenced this pull request Dec 10, 2025
Summary:

X-link: meta-pytorch/torchrec#3602

X-link: facebookresearch/FBGEMM#2202

Add output_dtype parameter to MX4 dequantization stack to support direct
conversion to BF16/FP16, avoiding expensive FP32 intermediate step.

Differential Revision: D87826479
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

CLA Signed This label is managed by the Facebook bot. Authors need to sign the CLA before a PR can be reviewed. fb-exported meta-exported

Projects

None yet

Development

Successfully merging this pull request may close these issues.

1 participant